Self-supervised and Few-shot Contrastive Learning Frameworks for Text Clustering
نویسندگان
چکیده
Contrastive learning is a promising approach to unsupervised learning, as it inherits the advantages of well-studied deep models without dedicated and complex model design. In this paper, based on bidirectional encoder representations from transformers (BERT) long-short term memory (LSTM) neural networks, we propose self-supervised contrastive (SCL) well few-shot (FCL) with data augmentation (UDA) for text clustering. BERT-SCL outperforms state-of-the-art clustering approaches short texts long in terms several evaluation measures. LSTM-SCL also shows good performance BERT-FCL achieves close supervised UDA further improves texts. LSTM-FCL Our experiment results suggest that both SCL FCL are effective
منابع مشابه
Meta-Learning for Semi-Supervised Few-Shot Classification
In few-shot classification, we are interested in learning algorithms that train a classifier from only a handful of labeled examples. Recent progress in few-shot classification has featured meta-learning, in which a parameterized model for a learning algorithm is defined and trained on episodes representing different classification problems, each with a small labeled training set and its corres...
متن کاملSemi-Supervised Learning for Web Text Clustering
Supervised learning algorithms usually require large amounts of training data to learn reasonably accurate classifiers. Yet, for many text classification tasks, providing labeled training documents is expensive, while unlabeled documents are readily available in large quantities. Learning from both, labeled and unlabeled documents, in a semi-supervised framework is a promising approach to reduc...
متن کاملSemi-Supervised Few-Shot Learning with Prototypical Networks
We consider the problem of semi-supervised few-shot classification (when the few labeled samples are accompanied with unlabeled data) and show how to adapt the Prototypical Networks [10] to this problem. We first show that using larger and better regularized prototypical networks can improve the classification accuracy. We then show further improvements by making use of unlabeled data.
متن کاملFew-shot Learning
Though deep neural networks have shown great success in the large data domain, they generally perform poorly on few-shot learning tasks, where a classifier has to quickly generalize after seeing very few examples from each class. The general belief is that gradient-based optimization in high capacity classifiers requires many iterative steps over many examples to perform well. Here, we propose ...
متن کاملTime-Contrastive Networks: Self-Supervised Learning from Video
We propose a self-supervised approach for learning representations and robotic behaviors entirely from unlabeled videos recorded from multiple viewpoints, and study how this representation can be used in two robotic imitation settings: imitating object interactions from videos of humans, and imitating human poses. Imitation of human behavior requires a viewpoint-invariant representation that ca...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2023
ISSN: ['2169-3536']
DOI: https://doi.org/10.1109/access.2023.3302913